Mixtures of Gaussians

نویسنده

  • Barbara Resch
چکیده

This tutorial treats mixtures of Gaussian probability distribution functions. Gaussian mixtures are combinations of a finite number of Gaussian distributions. They are used to model complex multidimensional distributions. When there is a need to learn the parameters of the Gaussian mixture, the EM algorithm is used. In the second part of this tutorial mixtures of Gaussian are used to model the emission probability distribution function in Hidden Markov Models. Usage To make full use of this tutorial you should 1. Download the file MixtGaussian.zip which contains this tutorial and the accompanying Matlab programs. 2. Unzip MixtGaussian.zip which will generate a subdirectory named MixtGaussian/matlab where you can find all the Matlab programs. 3. Add the folder MixtGaussian/matlab and the subfolders to the Matlab search path with a command like addpath(’C:\Work\MixtGaussian\matlab’) if you are using a Windows machine or addpath(’/home/jack/MixtGaussian/matlab’) if you are using a Unix/Linux machine. 1 Mixtures of Gaussians 1.1 Formulas and Definitions Gaussian Mixtures are combinations of Gaussian, or ‘normal’, distributions. A mixture of Gaussians can be written as a weighted sum of Gaussian densities. Recall the d-dimensional Gaussian probability density function (pdf): g(μ,Σ)(x) = 1 √ 2π d√ det (Σ) e− 1 2 (x−μ) TΣ−1(x−μ), (1) with mean vector μ and covariance matrix Σ. A weighted mixture of K Gaussians can be written as

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Learning Mixtures of Gaussians

Mixtures of Gaussians are among the most fundamental and widely used statistical models. Current techniques for learning such mixtures from data are local search heuristics with weak performance guarantees. We present the first provably correct algorithm for learning a mixture of Gaussians. This algorithm is very simple and returns the true centers of the Gaussians to within the precision speci...

متن کامل

EM for Spherical Gaussians

In this project, we examine two aspects of the behavior of the EM algorithm for mixtures of spherical Gaussians; 1) the benefit of spectral projection for such mixtures, and 2) the general behavior of the EM algorithm under certain separability criteria. Our current results are for mixtures of two Gaussians, although these can be extended. In the case of 1), we show that the value of the Q func...

متن کامل

PAC Learning Mixtures of Axis-Aligned Gaussians with No Separation Assumption

We propose and analyze a new vantage point for the learning of mixtures of Gaussians: namely, the PAC-style model of learning probability distributions introduced by Kearns et al. [13]. Here the task is to construct a hypothesis mixture of Gaussians that is statistically indistinguishable from the actual mixture generating the data; specifically, the KL divergence should be at most ǫ. In this s...

متن کامل

Design of compact acoustic models through clustering of tied-covariance Gaussians

We propose a new approach for designing compact acoustic models particularly suited to large systems that combine multiple model sets to represent distinct acoustic conditions or languages. We show that Gaussians based on mixtures of inverse covariances (MIC) with shared parameters can be clustered using an efficient Lloyd algorithm. As a result, more compact acoustic models can be built by clu...

متن کامل

PAC Learning Mixtures of Gaussians with No Separation Assumption

We propose and analyze a new vantage point for the learning of mixtures of Gaussians: namely, the PAC-style model of learning probability distributions introduced by Kearns et al. [12]. Here the task is to construct a hypothesis mixture of Gaussians that is statistically indistinguishable from the actual mixture generating the data; specifically, the KL divergence should be at most 2. In this s...

متن کامل

Tied Mixtures in the Lincoln Robust CSR

HMM recognizers using either a single Gaussian or a Gaussian mixture per state have been shown to work fairly well for 1000-word vocabulary continuous speech recognition. However, the large number of Gaussians required to cover the entire English language makes these systems unwieldy for large vocabulary tasks. Tied mixtures offer a more compact way of representing the observation pdf's. We hav...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2006